decomposition tree
Private Learning of Littlestone Classes, Revisited
We consider online and PAC learning of Littlestone classes subject to the constraint of approximate differential privacy. Our main result is a private learner to online-learn a Littlestone class with a mistake bound of $\tilde{O}(d^{9.5}\cdot \log(T))$ in the realizable case, where $d$ denotes the Littlestone dimension and $T$ the time horizon. This is a doubly-exponential improvement over the state-of-the-art [GL'21] and comes polynomially close to the lower bound for this task. The advancement is made possible by a couple of ingredients. The first is a clean and refined interpretation of the ``irreducibility'' technique from the state-of-the-art private PAC-learner for Littlestone classes [GGKM'21]. Our new perspective also allows us to improve the PAC-learner of [GGKM'21] and give a sample complexity upper bound of $\widetilde{O}(\frac{d^5 \log(1/δβ)}{\varepsilon α})$ where $α$ and $β$ denote the accuracy and confidence of the PAC learner, respectively. This improves over [GGKM'21] by factors of $\frac{d}α$ and attains an optimal dependence on $α$. Our algorithm uses a private sparse selection algorithm to \emph{sample} from a pool of strongly input-dependent candidates. However, unlike most previous uses of sparse selection algorithms, where one only cares about the utility of output, our algorithm requires understanding and manipulating the actual distribution from which an output is drawn. In the proof, we use a sparse version of the Exponential Mechanism from [GKM'21] which behaves nicely under our framework and is amenable to a very easy utility proof.
Online Sum-Product Computation over Trees
We consider the problem of performing efficient sum-product computations in an online setting over a tree. A natural application of our methods is to compute the marginal distribution at a vertex in a tree-structured Markov random field. Belief propagation can be used to solve this problem, but requires time linear in the size of the tree, and is therefore too slow in an online setting where we are continuously receiving new data and computing individual marginals. With our method we aim to update the data and compute marginals in time that is no more than logarithmic in the size of the tree, and is often significantly less. We accomplish this via a hierarchical covering structure that caches previous local sum-product computations. Our contribution is three-fold: we i) give a linear time algorithm to find an optimal hierarchical cover of a tree; ii) give a sum-productlike algorithm to efficiently compute marginals with respect to this cover; and iii) apply "i" and "ii" to find an efficient algorithm with a regret bound for the online allocation problem in a multi-task setting.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Europe > Italy > Lombardy > Milan (0.04)
Optimal sub-graphical models
The complexity of inference in graphical models is typically exponential in some parame- ter of the graph, such as the size of the largest clique. Therefore, it is often required to find a subgraphical model that has lower complexity (smaller clique size) without introducing a large error in inference results. The KL-divergence between the original probability dis- tribution and the probability distribution on the simplified graphical model is often used to measure the impact on inference. Existing techniques for reducing the complexity of graph- ical models including annihilation and edge-removal [4] are greedy in nature and cannot make any guarantees regarding the optimality of the solution. This problem is NP-complete [9] and so, in general, one cannot expect a polynomial time algorithm to find the optimal solution. However, we show that when we restrict the problem to some sets of subgraphs, the optimal solution can be found quite quickly using a dynamic programming algorithm in time polynomial in the tree-width of the graph.
Glue: Adaptively Merging Single Table Cardinality to Estimate Join Query Size
Zhu, Rong, Zeng, Tianjing, Pfadler, Andreas, Chen, Wei, Ding, Bolin, Zhou, Jingren
Cardinality estimation (CardEst), a central component of the query optimizer, plays a significant role in generating high-quality query plans in DBMS. The CardEst problem has been extensively studied in the last several decades, using both traditional and ML-enhanced methods. Whereas, the hardest problem in CardEst, i.e., how to estimate the join query size on multiple tables, has not been extensively solved. Current methods either reply on independence assumptions or apply techniques with heavy burden, whose performance is still far from satisfactory. Even worse, existing CardEst methods are often designed to optimize one goal, i.e., inference speed or estimation accuracy, which can not adapt to different occasions. In this paper, we propose a very general framework, called Glue, to tackle with these challenges. Its key idea is to elegantly decouple the correlations across different tables and losslessly merge single table CardEst results to estimate the join query size. Glue supports obtaining the single table-wise CardEst results using any existing CardEst method and can process any complex join schema. Therefore, it easily adapts to different scenarios having different performance requirements, i.e., OLTP with fast estimation time or OLAP with high estimation accuracy. Meanwhile, we show that Glue can be seamlessly integrated into the plan search process and is able to support counting distinct number of values. All these properties exhibit the potential advances of deploying Glue in real-world DBMS.
Power BI - AI and Q&A updates, decomposition tree, and data prep. (Microsoft Ignite)
Demonstration of the latest Power BI capabilities to help you to discover and explore insights in your data including: Automated Machine Learning for predictive modelling. This is Power BI's role as part of the Power Platform helping you to prepare your data across sources. At Microsoft Ignite 2019, this was session THR2285: Microsoft Power BI and the Power Platform: Dataflows, AI, and new visualizations. Justyna Lucznik is a Program Manager for the AI features of Power BI. She has worked on AI visualizations like the decomposition tree and key influencers.
- Information Technology > Software (1.00)
- Information Technology > Communications > Social Media (0.76)
- Information Technology > Artificial Intelligence > Machine Learning (0.70)
A Tour of Artificial Intelligence Features in Power BI
The Decomposition Tree is the latest AI visualization developed by the Power BI team. This is a highly interactive visual that allows you to decompose (break down) a measure by various attributes across different dimensions. This visual can be used for ad hoc, exploratory analysis to understand your data. Or, it can be used to perform root-cause analysis, which is enhanced by the built-in AI functionality. For example, when drilling into the tree, you can choose "High value" or "Low value".
- Information Technology > Artificial Intelligence (0.77)
- Information Technology > Software (0.68)
Refining HTN Methods via Task Insertion with Preferences
Xiao, Zhanhao, Wan, Hai, Zhuo, Hankui Hankz, Herzig, Andreas, Perrussel, Laurent, Chen, Peilin
Hierarchical Task Network (HTN) planning is showing its power in real-world planning. Although domain experts have partial hierarchical domain knowledge, it is time-consuming to specify all HTN methods, leaving them incomplete. On the other hand, traditional HTN learning approaches focus only on declarative goals, omitting the hierarchical domain knowledge. In this paper, we propose a novel learning framework to refine HTN methods via task insertion with completely preserving the original methods. As it is difficult to identify incomplete methods without designating declarative goals for compound tasks, we introduce the notion of prioritized preference to capture the incompleteness possibility of methods. Specifically, the framework first computes the preferred completion profile w.r .t.the prioritized preference to refine the incomplete methods. Then it finds the minimal set of refined methods via a method substitution operation. Experimental analysis demonstrates that our approach is effective, especially in solving new HTN planning instances.
This Is a Solution! (... But Is It Though?) - Verifying Solutions of Hierarchical Planning Problems
Behnke, Gregor (Ulm University) | Höller, Daniel (Ulm University) | Biundo, Susanne (Ulm University)
Plan-Verification is the task of determining whether a plan is a solution to a given planning problem. Any plan verifier has, apart from showing that verifying plans is possible in practice, a wide range of possible applications. These include mixed-initiative planning, where a user is integrated into the planning process, and local search, e.g., for post-optimising plans or for plan repair. In addition to its practical interest, plan verification is also a problem worth investigating for theoretical reasons. Recent work showed plan verification for hierarchical planning problems to be NP-complete, as opposed to classical planning where it is in P. As such, plan verification for hierarchical planning problem was — until now — not possible. We describe the first plan verifier for hierarchical planning. It uses a translation of the problem into a SAT formula. Further we conduct an empirical evaluation, showing that the correct output is produced within acceptable time.
Active Preference Elicitation for Planning
Das, Mayukh (Indiana University Bloomington) | Islam, Md. Rakibul (Washington State University) | Doppa, Janardhan Rao (Jana) (Washington State University) | Roth, Dan (University of Illinois at Urbana-Champaign) | Natarajan, Sriraam (Indiana University Bloomington)
We consider the problem of actively eliciting preferences from a human by a planning system. While prior work in planning have explored the use of domain knowledge and preferences, they assume that the knowledge must be provided before the planner starts the planning process. Our work is in building more collaborative systems where a system can solicit advice as needed. We verify empirically that this approach lead to faster and better solutions, while reducing the burden on the human expert.
- North America > United States > Washington > Whitman County > Pullman (0.05)
- North America > United States > Indiana > Monroe County > Bloomington (0.05)
- North America > United States > Oklahoma > Payne County > Cushing (0.04)
- North America > United States > Illinois > Champaign County > Urbana (0.04)
Deciding Monotone Duality and Identifying Frequent Itemsets in Quadratic Logspace
The monotone duality problem is defined as follows: Given two monotone formulas f and g in iredundant DNF, decide whether f and g are dual. This problem is the same as duality testing for hypergraphs, that is, checking whether a hypergraph H consists of precisely all minimal transversals of a simple hypergraph G. By exploiting a recent problem-decomposition method by Boros and Makino (ICALP 2009), we show that duality testing for hypergraphs, and thus for monotone DNFs, is feasible in DSPACE[log^2 n], i.e., in quadratic logspace. As the monotone duality problem is equivalent to a number of problems in the areas of databases, data mining, and knowledge discovery, the results presented here yield new complexity results for those problems, too. For example, it follows from our results that whenever for a Boolean-valued relation (whose attributes represent items), a number of maximal frequent itemsets and a number of minimal infrequent itemsets are known, then it can be decided in quadratic logspace whether there exist additional frequent or infrequent itemsets.
- North America > United States > New York > New York County > New York City (0.04)
- Asia > China > Hong Kong (0.04)